More AD of Nonlinear AMPL Models: Computing Hessian Information and Exploiting Partial Separability†
نویسنده
چکیده
We describe computational experience with automatic differentiation of mathematical programming problems expressed in the modeling language AMPL. Nonlinear expressions are translated to loop-free code, which makes it easy to compute gradients and Jacobians by backward automatic differentiation. The nonlinear expressions may be interpreted or, to gain some evaluation speed at the cost of increased preparation time, converted to Fortran or C. We have extended the interpretive scheme to evaluate Hessian (of Lagrangian) times vector. Detecting partially separable structure (sums of terms, each depending, perhaps after a linear transformation, on only a few variables) is of independent interest, as some solvers exploit this structure. It can be detected automatically by suitable ‘‘tree walks’’. Exploiting this structure permits an AD computation of the entire Hessian matrix by accumulating Hessian times vector computations for each term, and can lead to a much faster computation of the Hessian than by computing the whole Hessian times each unit vector.
منابع مشابه
Automatically Finding and Exploiting Partially Separable Structure in Nonlinear Programming Problems
Nonlinear programming problems often involve an objective and constraints that are partially separable — the sum of terms involving only a few variables (perhaps after a linear change of variables). This paper discusses finding and exploiting such structure in nonlinear programming problems expressed symbolically in the AMPL modeling language. For some computations, such as computing Hessians b...
متن کاملA New Concept for Separability Problems in Blind Source Separation
The goal of blind source separation (BSS) lies in recovering the original independent sources of a mixed random vector without knowing the mixing structure. A key ingredient for performing BSS successfully is to know the indeterminacies of the problem-that is, to know how the separating model relates to the original mixing model (separability). For linear BSS, Comon (1994) showed using the Darm...
متن کاملEfficient (Partial) Determination of Derivative Matrices via Automatic Differentiation
In many scientific computing applications involving nonlinear systems or methods of optimization, a sequence of Jacobian or Hessian matrices is required. Automatic differentiation (AD) technology can be used to accurately determine these matrices, and it is well known that if these matrices exhibit a sparsity pattern (for all iterates), then not only can AD take advantage of this sparsity for s...
متن کاملMINLPLib - A Collection of Test Models for Mixed-Integer Nonlinear Programming
The paper describes a new computerized collection of test models for mixed-integer nonlinear programming. Since there is no standard format for nonlinear models, the model collection is augmented with a translation server that can transform the models from their basic GAMS format into a number of other formats such as AMPL, BARON, LINGO, and MINOPT. The translation server can also be used to tr...
متن کاملExploiting separability in large-scale Support Vector Machine training
Support vector machine training can be represented as a large quadratic program. We present an efficient and numerically stable algorithm for this problem using interior point methods, which requires only O(n) operations per iteration. Through exploiting the separability of the Hessian, we provide a unified approach, from an optimization perspective, to 1-norm classification, 2-norm classificat...
متن کامل